Convolutional Neural Networks With Dynamic Regularization
نویسندگان
چکیده
Regularization is commonly used for alleviating overfitting in machine learning. For convolutional neural networks (CNNs), regularization methods, such as DropBlock and Shake-Shake, have illustrated the improvement generalization performance. However, these methods lack a self-adaptive ability throughout training. That is, strength fixed to predefined schedule, manual adjustments are required adapt various network architectures. In this article, we propose dynamic method CNNs. Specifically, model function of training loss. According change loss, our can dynamically adjust procedure, thereby balancing underfitting With regularization, large-scale automatically regularized by strong perturbation, vice versa. Experimental results show that proposed improve capability on off-the-shelf architectures outperform state-of-the-art methods.
منابع مشابه
Convolutional neural networks with low-rank regularization
Large CNNs have delivered impressive performance in various computer vision applications. But the storage and computation requirements make it problematic for deploying these models on mobile devices. Recently, tensor decompositions have been used for speeding up CNNs. In this paper, we further develop the tensor decomposition technique. We propose a new algorithm for computing the low-rank ten...
متن کاملImproved Regularization of Convolutional Neural Networks with Cutout
Convolutional neural networks are capable of learning powerful representational spaces, which are necessary for tackling complex learning tasks. However, due to the model capacity required to capture such representations, they are often susceptible to overfitting and therefore require proper regularization in order to generalize well. In this paper, we show that the simple regularization techni...
متن کاملDynamic Sampling Convolutional Neural Networks
We present Dynamic Sampling Convolutional Neural Networks (DSCNN), where the position-specific kernels learn from not only the current position but also multiple sampled neighbour regions. During sampling, residual learning is introduced to ease training and an attention mechanism is applied to fuse features from different samples. And the kernels are further factorized to reduce parameters. Th...
متن کاملMax-Pooling Dropout for Regularization of Convolutional Neural Networks
Recently, dropout has seen increasing use in deep learning. For deep convolutional neural networks, dropout is known to work well in fully-connected layers. However, its effect in pooling layers is still not clear. This paper demonstrates that max-pooling dropout is equivalent to randomly picking activation based on a multinomial distribution at training time. In light of this insight, we advoc...
متن کاملStochastic Pooling for Regularization of Deep Convolutional Neural Networks
We introduce a simple and effective method for regularizing large convolutional neural networks. We replace the conventional deterministic pooling operations with a stochastic procedure, randomly picking the activation within each pooling region according to a multinomial distribution, given by the activities within the pooling region. The approach is hyper-parameter free and can be combined wi...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE transactions on neural networks and learning systems
سال: 2021
ISSN: ['2162-237X', '2162-2388']
DOI: https://doi.org/10.1109/tnnls.2020.2997044